Gradient-Variation Bound for Online Convex Optimization with Constraints

نویسندگان

چکیده

We study online convex optimization with constraints consisting of multiple functional and a relatively simple constraint set, such as Euclidean ball. As enforcing the at each time step through projections is computationally challenging in general, we allow decisions to violate but aim achieve low regret cumulative violation over horizon T steps. First-order methods an O(sqrt{T}) O(1) violation, which best-known bound under Slater's condition, do not take into account structural information problem. Furthermore, existing algorithms analysis are limited space. In this paper, provide instance-dependent for complex obtained by novel primal-dual mirror-prox algorithm. Our quantified total gradient variation V_*(T) sequence loss functions. The proposed algorithm works general normed spaces simultaneously achieves O(sqrt{V_*(T)}) never worse than (O(sqrt{T}), O(1)) result improves previous that applied mirror-prox-type problem achieving O(T^{2/3}) violation. Finally, our efficient, it only performs mirror descent steps iteration instead solving Lagrangian minimization

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Regret Bound by Variation for Online Convex Optimization

In (Hazan and Kale, 2008), the authors showed that the regret of the Follow the Regularized Leader (FTRL) algorithm for online linear optimization can be bounded by the total variation of the cost vectors. In this paper, we extend this result to general online convex optimization. We first analyze the limitations of the FTRL algorithm in (Hazan and Kale, 2008) when applied to online convex opti...

متن کامل

Adaptive Bound Optimization for Online Convex Optimization

We introduce a new online convex optimization algorithm that adaptively chooses its regularization function based on the loss functions observed so far. This is in contrast to previous algorithms that use a fixed regularization function such as L2-squared, and modify it only via a single time-dependent parameter. Our algorithm’s regret bounds are worst-case optimal, and for certain realistic cl...

متن کامل

Online Convex Optimization with Stochastic Constraints

This paper considers online convex optimization (OCO) with stochastic constraints, which generalizes Zinkevich’s OCO over a known simple fixed set by introducing multiple stochastic functional constraints that are i.i.d. generated at each round and are disclosed to the decision maker only after the decision is made. This formulation arises naturally when decisions are restricted by stochastic e...

متن کامل

Online Convex Optimization for Cumulative Constraints

We propose an algorithm for online convex optimization which examines a clipped long-term constraint of the form T ∑ t=1 [g(xt)]+, encoding the cumulative constraint violation. Previous literature has focused on longterm constraints of the form T ∑ t=1 g(xt), for which strictly feasible solutions can cancel out the effects of violated constraints. In this paper, we generalize the results in [16...

متن کامل

Adaptive Algorithms for Online Convex Optimization with Long-term Constraints

We present an adaptive online gradient descent algorithm to solve online convex optimization problems with long-term constraints, which are constraints that need to be satisfied when accumulated over a finite number of rounds T , but can be violated in intermediate rounds. For some user-defined trade-off parameter β ∈ (0, 1), the proposed algorithm achieves cumulative regret bounds of O(Tmax{β,...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2023

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v37i8.26141